Discontinuous Generalization in Large Committee Machines

نویسندگان

  • Holm Schwarze
  • John A. Hertz
چکیده

J. Hertz Nordita Blegdamsvej 17 2100 Copenhagen 0 Denmark The problem of learning from examples in multilayer networks is studied within the framework of statistical mechanics. Using the replica formalism we calculate the average generalization error of a fully connected committee machine in the limit of a large number of hidden units. If the number of training examples is proportional to the number of inputs in the network, the generalization error as a function of the training set size approaches a finite value. If the number of training examples is proportional to the number of weights in the network we find first-order phase transitions with a discontinuous drop in the generalization error for both binary and continuous weights.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Statistical Mechanics of Learning in a Large Committee Machine

In a We use statistical mechanics to study generalization in large committee machines. For an architecture with nonoverlapping receptive fields a replica calculation yields the generalization error in the limit of a large number of hidden units. For continuous weights the generalization error falls off asymptotically inversely proportional to Q, the number of training examples per weight. For b...

متن کامل

Learning in ultrametric committee machines

Abstract The problem of learning by examples in ultrametric committee machines (UCMs) is studied within the framework of statistical mechanics. Using the replica formalism we calculate the average generalization error in UCMs with L hidden layers and for a large enough number of units. In most of the regimes studied we find that the generalization error, as a function of the number of examples ...

متن کامل

Generalization in a Large Committee Machine

We study generalization in a committee machine with non{overlapping receptive elds trained to implement a function of the same structure. Using the replica method, we calculate the generalization error in the limit of a large number of hidden units. For continuous weights the generalization error falls oo asymptotically inversely proportional to , the number of training examples per weight. For...

متن کامل

ar X iv : c on d - m at / 9 60 11 22 v 1 2 5 Ja n 19 96 Learning and Generalization Theories of Large Committee – Machines

The study of the distribution of volumes associated to the internal representations of learning examples allows us to derive the critical learning capacity (αc = 16 π √ lnK) of large committee machines, to verify the stability of the solution in the limit of a large number K of hidden units and to find a Bayesian generalization cross–over at α = K. PACS Numbers : 05.20 64.60 87.10 Typeset using...

متن کامل

Universal Asymptotics in Committee Machines with Tree Architecture

On-line supervised learning in the general K Tree Committee Machine (TCM) is studied for a uniform distribution of inputs. Examples are corrupted by noise in the teacher output. From the diierential equations which describe the learning dynamics, the algorithm which optimizes the generalization ability is exactly obtained. For a large number of presented examples, the asymptotical behaviour of ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1993